13 research outputs found

    A network flow algorithm for just-in-time project scheduling

    Get PDF
    We show the polynomial solvability of the PERT-COST project scheduling problem in the case of: (i) the objective being a piecewise-linear, convex (possibly, non- monotone) function of the job durations as well as of job start/finish times, and (ii) the precedence relations between jobs being presented in the form of a general (not necessary, acyclic) directed graph with arc lengths of any sign. For the latter problem, we present a network flow algorithm (of pseudo-linear complexity) which is easy to implement and which behaves well when the objective values grow slowly with the growth of the problem size while the number of breakpoints in the objective grows fast

    Implementation of an Optimal First-Order Method for Strongly Convex Total Variation Regularization

    Get PDF
    We present a practical implementation of an optimal first-order method, due to Nesterov, for large-scale total variation regularization in tomographic reconstruction, image deblurring, etc. The algorithm applies to μ\mu-strongly convex objective functions with LL-Lipschitz continuous gradient. In the framework of Nesterov both μ\mu and LL are assumed known -- an assumption that is seldom satisfied in practice. We propose to incorporate mechanisms to estimate locally sufficient μ\mu and LL during the iterations. The mechanisms also allow for the application to non-strongly convex functions. We discuss the iteration complexity of several first-order methods, including the proposed algorithm, and we use a 3D tomography problem to compare the performance of these methods. The results show that for ill-conditioned problems solved to high accuracy, the proposed method significantly outperforms state-of-the-art first-order methods, as also suggested by theoretical results.Comment: 23 pages, 4 figure

    Towards Machine Wald

    Get PDF
    The past century has seen a steady increase in the need of estimating and predicting complex systems and making (possibly critical) decisions with limited information. Although computers have made possible the numerical evaluation of sophisticated statistical models, these models are still designed \emph{by humans} because there is currently no known recipe or algorithm for dividing the design of a statistical model into a sequence of arithmetic operations. Indeed enabling computers to \emph{think} as \emph{humans} have the ability to do when faced with uncertainty is challenging in several major ways: (1) Finding optimal statistical models remains to be formulated as a well posed problem when information on the system of interest is incomplete and comes in the form of a complex combination of sample data, partial knowledge of constitutive relations and a limited description of the distribution of input random variables. (2) The space of admissible scenarios along with the space of relevant information, assumptions, and/or beliefs, tend to be infinite dimensional, whereas calculus on a computer is necessarily discrete and finite. With this purpose, this paper explores the foundations of a rigorous framework for the scientific computation of optimal statistical estimators/models and reviews their connections with Decision Theory, Machine Learning, Bayesian Inference, Stochastic Optimization, Robust Optimization, Optimal Uncertainty Quantification and Information Based Complexity.Comment: 37 page

    A network flow algorithm for just-in-time project scheduling

    No full text
    We consider a project scheduling problem with the objective being a piecewise-linear, convex (possibly, non-monotone) function of the job durations as well as of job start/finish times. A version of ‘out-of-kilter’ algorithm of pseudo-linear complexity to handle this problem is provided

    An optimal ( ϵ

    No full text
    corecore